Developing Population Codes by Minimizing Description Length

نویسندگان

  • Richard S. Zemel
  • Geoffrey E. Hinton
چکیده

The Minimum Description Length principle (MDL) can be used to train the hidden units of a neural network to extract a representation that is cheap to describe but nonetheless allows the input to be reconstructed accurately. We show how MDL can be used to develop highly redundant population codes. Each hidden unit has a location in a lowdimensional implicit space. If the hidden unit activities form a bump of a standard shape in this space, they can be cheaply encoded by the center of this bump. So the weights from the input units to the hidden units in a self-supervised network are trained to make the activities form a standard bump. The coordinates of the hidden units in the implicit space are also learned, thus allowing exibility, as the network develops a discontinuous topography when presented with di erent input classes. Population-coding in a space other than the input enables a network to extract nonlinear higher-order properties of the inputs. Most existing unsupervised learning algorithms can be understood using the Minimum Description Length (MDL) principle (Rissanen, 1989). Given an ensemble of input vectors, the aim of the learning algorithm is to nd a method of coding each input vector that minimizes the total cost, in bits, of communicating the input vectors to a receiver. There are three terms in the total description length: The code-cost is the number of bits required to communicate the code that the algorithm assigns to each input vector. The model-cost is the number of bits required to specify how to reconstruct input vectors from codes (e.g., the hidden-to-output weights in Figure 1). 1Corresponding author. Algorithms and Architectures: Learning Algorithms|Oral

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Learning Population Codes by Minimizing Description Length

The Minimum Description Length principle (MDL) can be used to train the hidden units of a neural network to extract a representation that is cheap to describe but nonetheless allows the input to be reconstructed accurately. We show how MDL can be used to develop highly redundant population codes. Each hidden unit has a location in a low-dimensional implicit space. If the hidden unit activities ...

متن کامل

Developing Population Codes For Object Instantiation Parameters

An efficient and useful representation for an object viewed from different positions is in terms of its instantiation parameters. We show how the Minimum Description Length principle (MDL) can be used to train the hidden units of a neural network to develop a population code for the instantiation parameters of an object in an image. Each hidden unit has a location in a low-dimensional implicit ...

متن کامل

Compact Genetic Codes as a Search Strategy of Evolutionary Processes

The choice of genetic representation crucially determines the capability of evolutionary processes to find complex solutions in which many variables interact. The question is how good genetic representations can be found and how they can be adapted online to account for what can be learned about the structure of the problem from previous samples. We address these questions in a scenario that we...

متن کامل

Developing the Job Stress Model for Airlines Flight crew

Job is one of the most important issues in the discussion of human resources, and since the discussion of job stress in the flight crew and its relationship with the performance of the flight crew is very important and vital. This research is about recognizing and developing a model for job stress among the cabin crew of Mahan Airline. The qualitative research methodology has been followed by q...

متن کامل

Universal coding with minimum probability of codeword length overflow

-Lossless block-to-variable length source coding is studied for finite-state, finite-alphabet sources. We aim to minimize the probability that the normalized length of the codeword will exceed a given threshold B, subject to the Kraft inequality. It is shown that the Lempel-Ziv (LZ) algorithm asymptotically attains the optimal performance in the sense just defined, independently of the source a...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 1993